Automatic Capacity Tuning of Very Large VC-Dimension Classifiers
نویسندگان
چکیده
Large VC-dimension classifiers can learn difficult tasks, but are usually impractical because they generalize well only if they are trained with huge quantities of data. In this paper we show that even high-order polynomial classifiers in high dimensional spaces can be trained with a small amount of training data and yet generalize better than classifiers with a smaller VC-dimension. This is achieved with a maximum margin algorithm (the Generalized Portrait). The technique is applicable to a wide variety of classifiers, including Perceptrons, polynomial classifiers (sigma-pi unit networks) and Radial Basis Functions. The effective number of parameters is adjusted automatically by the training algorithm to match the complexity of the problem. It is shown to equal the number of those training patterns which are closest patterns to the decision boundary (supporting patterns). Bounds on the generalization error and the speed of convergence of the algorithm are given. Experimental results on handwritten digit recognition demonstrate good generalization compared to other algorithms.
منابع مشابه
Automatic Capacity Tuning of Very Large Vc-dimension Classiers
Large VC-dimension classiers can learn dicult tasks, but are usually impractical because they generalize well only if they are trained with huge quantities of data. In this paper we show that even very high-order polynomial classiers can be trained with a small amount of training data and yet generalize better than classiers with a smaller VC-dimension. This is achieved with a maximum margin al...
متن کاملScale-sensitive Ψ -dimensions: the Capacity Measures for Classifiers Taking Values in R
Bounds on the risk play a crucial role in statistical learning theory. They usually involve as capacity measure of the model studied the VC dimension or one of its extensions. In classification, such “VC dimensions” exist for models taking values in {0, 1}, {1, . . . , Q} and R. We introduce the generalizations appropriate for the missing case, the one of models with values in R. This provides ...
متن کاملThe VC-Dimension versus the Statistical Capacity of Multilayer Networks
A general relationship is developed between the VC-dimension and the statistical lower epsilon-capacity which shows that the VC-dimension can be lower bounded (in order) by the statistical lower epsilon-capacity of a network trained with random samples. This relationship explains quantitatively how generalization takes place after memorization, and relates the concept of generalization (consist...
متن کاملha l - 00 15 69 14 , v er si on 1 - 2 5 Ju n 20 07 Scale - sensitive Ψ - dimensions : the Capacity Measures for Classifiers Taking Values in R Q
Bounds on the risk play a crucial role in statistical learning theory. They usually involve as capacity measure of the model studied the VC dimension or one of its extensions. In classification, such “VC dimensions” exist for models taking values in {0, 1}, {1, . . . , Q} and R. We introduce the generalizations appropriate for the missing case, the one of models with values in R. This provides ...
متن کاملScale-sensitive Psi-dimensions: the Capacity Measures for Classifiers Taking Values in R^Q
Bounds on the risk play a crucial role in statistical learning theory. They usually involve as capacity measure of the model studied the VC dimension or one of its extensions. In classification, such “VC dimensions” exist for models taking values in {0, 1}, {1, . . . , Q} and R. We introduce the generalizations appropriate for the missing case, the one of models with values in R. This provides ...
متن کامل